The Kullback-Leibler (KL) divergence is a distance-like measure between the distribution of two random variables. It is defined as
It is not symmetric, so it is not a true measure of distance.
Derivation of the KL divergence
Recall that cross-entropy is defined as
When
Hence cross-entropy is not a distance measure between the distributions
Applying the logarithm property that